Gradient-Variation Bound for Online Convex Optimization with Constraints
نویسندگان
چکیده
We study online convex optimization with constraints consisting of multiple functional and a relatively simple constraint set, such as Euclidean ball. As enforcing the at each time step through projections is computationally challenging in general, we allow decisions to violate but aim achieve low regret cumulative violation over horizon T steps. First-order methods an O(sqrt{T}) O(1) violation, which best-known bound under Slater's condition, do not take into account structural information problem. Furthermore, existing algorithms analysis are limited space. In this paper, provide instance-dependent for complex obtained by novel primal-dual mirror-prox algorithm. Our quantified total gradient variation V_*(T) sequence loss functions. The proposed algorithm works general normed spaces simultaneously achieves O(sqrt{V_*(T)}) never worse than (O(sqrt{T}), O(1)) result improves previous that applied mirror-prox-type problem achieving O(T^{2/3}) violation. Finally, our efficient, it only performs mirror descent steps iteration instead solving Lagrangian minimization
منابع مشابه
Regret Bound by Variation for Online Convex Optimization
In (Hazan and Kale, 2008), the authors showed that the regret of the Follow the Regularized Leader (FTRL) algorithm for online linear optimization can be bounded by the total variation of the cost vectors. In this paper, we extend this result to general online convex optimization. We first analyze the limitations of the FTRL algorithm in (Hazan and Kale, 2008) when applied to online convex opti...
متن کاملAdaptive Bound Optimization for Online Convex Optimization
We introduce a new online convex optimization algorithm that adaptively chooses its regularization function based on the loss functions observed so far. This is in contrast to previous algorithms that use a fixed regularization function such as L2-squared, and modify it only via a single time-dependent parameter. Our algorithm’s regret bounds are worst-case optimal, and for certain realistic cl...
متن کاملOnline Convex Optimization with Stochastic Constraints
This paper considers online convex optimization (OCO) with stochastic constraints, which generalizes Zinkevich’s OCO over a known simple fixed set by introducing multiple stochastic functional constraints that are i.i.d. generated at each round and are disclosed to the decision maker only after the decision is made. This formulation arises naturally when decisions are restricted by stochastic e...
متن کاملOnline Convex Optimization for Cumulative Constraints
We propose an algorithm for online convex optimization which examines a clipped long-term constraint of the form T ∑ t=1 [g(xt)]+, encoding the cumulative constraint violation. Previous literature has focused on longterm constraints of the form T ∑ t=1 g(xt), for which strictly feasible solutions can cancel out the effects of violated constraints. In this paper, we generalize the results in [16...
متن کاملAdaptive Algorithms for Online Convex Optimization with Long-term Constraints
We present an adaptive online gradient descent algorithm to solve online convex optimization problems with long-term constraints, which are constraints that need to be satisfied when accumulated over a finite number of rounds T , but can be violated in intermediate rounds. For some user-defined trade-off parameter β ∈ (0, 1), the proposed algorithm achieves cumulative regret bounds of O(Tmax{β,...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2023
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v37i8.26141